Skip to content

Conversation

@chai-xiaonan
Copy link
Contributor

The checkpoint conversion tool must be updated accordingly following the Megatron update.
Error occurred when converting the trained torch-format ckpt to HF safetensor format:
1)The parameter args.expert_tensor_parallel_sizecannot be retrieved.
2)The model_provideris missing the model_builderparameter.
Fixed the two issues mentioned above.

@CLAassistant
Copy link

CLAassistant commented Nov 27, 2025

CLA assistant check
All committers have signed the CLA.

@lxd-cumt lxd-cumt changed the title update_conversion_tool [Train] Sync checkpoint conversion tools to the newest third_patry/Megatron-LM Dec 1, 2025
@lxd-cumt lxd-cumt changed the title [Train] Sync checkpoint conversion tools to the newest third_patry/Megatron-LM [Train] Sync checkpoint conversion tools to the newest third_party/Megatron-LM Dec 1, 2025
…L file, and resolved the absence of the tokenizer_typeparameter in the Aquila model conversion tool.
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants